Goto

Collaborating Authors

 Stuttgart



Elephants are smart. So are their whiskers.

Popular Science

Environment Animals Wildlife Elephants are smart. Their 1,000 whiskers make them dextrous enough to pick up a tortilla chip. Breakthroughs, discoveries, and DIY tips sent six days a week. An elephant's trunk is a wonder of evolution. Gentle, yet dextrous, it can pick up solid items, help them communicate, and be a helpful showering too l.




Multi-context principal component analysis

Wang, Kexin, Bhate, Salil, Pereira, João M., Kileel, Joe, Figlerowicz, Matylda, Seigal, Anna

arXiv.org Machine Learning

Principal component analysis (PCA) is a tool to capture factors that explain variation in data. Across domains, data are now collected across multiple contexts (for example, individuals with different diseases, cells of different types, or words across texts). While the factors explaining variation in data are undoubtedly shared across subsets of contexts, no tools currently exist to systematically recover such factors. We develop multi-context principal component analysis (MCPCA), a theoretical and algorithmic framework that decomposes data into factors shared across subsets of contexts. Applied to gene expression, MCPCA reveals axes of variation shared across subsets of cancer types and an axis whose variability in tumor cells, but not mean, is associated with lung cancer progression. Applied to contextualized word embeddings from language models, MCPCA maps stages of a debate on human nature, revealing a discussion between science and fiction over decades. These axes are not found by combining data across contexts or by restricting to individual contexts. MCPCA is a principled generalization of PCA to address the challenge of understanding factors underlying data across contexts.


Never Out of Date: How Hannah Arendt Helps Us Understand Our World

Der Spiegel International

Fifty years after her death in New York, Hannah Arendt has become the most popular philosopher of our time. For good reason: Her views are just as timely as ever. It must be so nice to play Hannah Arendt. No fewer than five actresses are on stage this evening at the Deutsches Theater Berlin to portray the philosopher. The piece is an adaptation of the graphic novel by American illustrator Ken Krimstein about the philosopher's life, called The Three Escapes of Hannah Arendt," combined with scenes from the famous interview that journalist Günter Gaus conducted with Arendt in 1964 for German public broadcaster ZDF. The article you are reading originally appeared in German in issue 49/2025 (November 28th, 2025) of DER SPIEGEL. They play Arendt and a few of her contemporaries, the philosopher Martin Heidegger, the writer Walter Benjamin, her husband Heinrich Blücher. There is a great deal of speech in the play, especially from Arendt herself. The places of her life are ticked off, her ...


GraphBench: Next-generation graph learning benchmarking

Stoll, Timo, Qian, Chendi, Finkelshtein, Ben, Parviz, Ali, Weber, Darius, Frasca, Fabrizio, Shavit, Hadar, Siraudin, Antoine, Mielke, Arman, Anastacio, Marie, Müller, Erik, Bechler-Speicher, Maya, Bronstein, Michael, Galkin, Mikhail, Hoos, Holger, Niepert, Mathias, Perozzi, Bryan, Tönshoff, Jan, Morris, Christopher

arXiv.org Machine Learning

Machine learning on graphs has recently achieved impressive progress in various domains, including molecular property prediction and chip design. However, benchmarking practices remain fragmented, often relying on narrow, task-specific datasets and inconsistent evaluation protocols, which hampers reproducibility and broader progress. To address this, we introduce GraphBench, a comprehensive benchmarking suite that spans diverse domains and prediction tasks, including node-level, edge-level, graph-level, and generative settings. GraphBench provides standardized evaluation protocols -- with consistent dataset splits and performance metrics that account for out-of-distribution generalization -- as well as a unified hyperparameter tuning framework. Additionally, we benchmark GraphBench using message-passing neural networks and graph transformer models, providing principled baselines and establishing a reference performance. See www.graphbench.io for further details.


From Many Models, One: Macroeconomic Forecasting with Reservoir Ensembles

Ballarin, Giovanni, Grigoryeva, Lyudmila, Li, Yui Ching

arXiv.org Machine Learning

Model combination is a powerful approach to achieve superior performance with a set of models than by just selecting any single one. We study both theoretically and empirically the effectiveness of ensembles of Multi-Frequency Echo State Networks (MFESNs), which have been shown to achieve state-of-the-art macroeconomic time series forecasting results (Ballarin et al., 2024a). Hedge and Follow-the-Leader schemes are discussed, and their online learning guarantees are extended to the case of dependent data. In applications, our proposed Ensemble Echo State Networks show significantly improved predictive performance compared to individual MFESN models.


Learning (Approximately) Equivariant Networks via Constrained Optimization

Manolache, Andrei, Chamon, Luiz F. O., Niepert, Mathias

arXiv.org Artificial Intelligence

Equivariant neural networks are designed to respect symmetries through their architecture, boosting generalization and sample efficiency when those symmetries are present in the data distribution. Real-world data, however, often departs from perfect symmetry because of noise, structural variation, measurement bias, or other symmetry-breaking effects. Strictly equivariant models may struggle to fit the data, while unconstrained models lack a principled way to leverage partial symmetries. Even when the data is fully symmetric, enforcing equivariance can hurt training by limiting the model to a restricted region of the parameter space. Guided by homotopy principles, where an optimization problem is solved by gradually transforming a simpler problem into a complex one, we introduce Adaptive Constrained Equivariance (ACE), a constrained optimization approach that starts with a flexible, non-equivariant model and gradually reduces its deviation from equivariance. This gradual tightening smooths training early on and settles the model at a data-driven equilibrium, balancing between equivariance and non-equivariance. Across multiple architectures and tasks, our method consistently improves performance metrics, sample efficiency, and robustness to input perturbations compared with strictly equivariant models and heuristic equivariance relaxations.


Systematic Framework of Application Methods for Large Language Models in Language Sciences

Sun, Kun, Wang, Rong

arXiv.org Artificial Intelligence

Large Language Models (LLMs) are transforming language sciences. However, their widespread deployment currently suffers from methodological fragmentation and a lack of systematic soundness. This study proposes two comprehensive methodological frameworks designed to guide the strategic and responsible application of LLMs in language sciences. The first method-selection framework defines and systematizes three distinct, complementary approaches, each linked to a specific research goal: (1) prompt-based interaction with general-use models for exploratory analysis and hypothesis generation; (2) fine-tuning of open-source models for confirmatory, theory-driven investigation and high-quality data generation; and (3) extraction of contextualized embeddings for further quantitative analysis and probing of model internal mechanisms. We detail the technical implementation and inherent trade-offs of each method, supported by empirical case studies. Based on the method-selection framework, the second systematic framework proposed provides constructed configurations that guide the practical implementation of multi-stage research pipelines based on these approaches. We then conducted a series of empirical experiments to validate our proposed framework, employing retrospective analysis, prospective application, and an expert evaluation survey. By enforcing the strategic alignment of research questions with the appropriate LLM methodology, the frameworks enable a critical paradigm shift in language science research. We believe that this system is fundamental for ensuring reproducibility, facilitating the critical evaluation of LLM mechanisms, and providing the structure necessary to move traditional linguistics from ad-hoc utility to verifiable, robust science.